Optimized Nearest Neighbor Methods Cam Weighted Distance & Statistical Confidence
نویسنده
چکیده
Nearest neighbor classification methods are a useful and a relatively straightforward to implement classification technique. However, despite such appeal, they still suffer from the curse of dimensionality. Additionally, the nature of the data sets may not be wholly applicable to the model assumed in the nearest neighbor methods. As such there have been many proposed optimizations. Two such optimizations studied in this project are statistical confidence and “cam-weighted” distance measure. For statistical confidence, a confidence measure is used to adapt the k value for k nearest neighbor to allow a more optimal set of neighbors for classification. Cam-weighted distance mimics attractive and repulsive forces of prototypes in determining its non-metric distance measure. This report documents the progress of studying these methods and their interaction.
منابع مشابه
Improving nearest neighbor classification with cam weighted distance
Nearest neighbor (NN) classification assumes locally constant class conditional probabilities, and suffers from bias in high dimensions with a small sample set. In this paper, we propose a novel cam weighted distance to ameliorate the curse of dimensionality. Different from the existing neighborhood-based methods which only analyze a small space emanating from the query sample, the proposed nea...
متن کاملBounds on the Power-weighted Mean Nearest Neighbor Distance
In this paper, bounds on the mean power-weighted nearest neighbor distance are derived. Previous work concentrates mainly on the infinite sample limit, whereas our bounds hold for any sample size. The results are expected to be of importance for example in statistical physics, nonparametric statistics and computational geometry, where they are related to the structure of matter as well as prope...
متن کاملA Statistical Confidence-Based Adaptive Nearest Neighbor Algorithm for Pattern Classification
The k-nearest neighbor rule is one of the simplest and most attractive pattern classification algorithms. It can be interpreted as an empirical Bayes classifier based on the estimated a posteriori probabilities from the k nearest neighbors. The performance of the k-nearest neighbor rule relies on the locally constant a posteriori probability assumption. This assumption, however, becomes problem...
متن کاملA New Distance-weighted k-nearest Neighbor Classifier
In this paper, we develop a novel Distance-weighted k -nearest Neighbor rule (DWKNN), using the dual distance-weighted function. The proposed DWKNN is motivated by the sensitivity problem of the selection of the neighborhood size k that exists in k -nearest Neighbor rule (KNN), with the aim of improving classification performance. The experiment results on twelve real data sets demonstrate that...
متن کاملMulti-hypothesis nearest-neighbor classifier based on class-conditional weighted distance metric
The performance of nearest-neighbor (NN) classifiers is known to be very sensitive to the distance metric used in classifying a query pattern, especially in scarce-prototype cases. In this paper, a classconditional weighted (CCW) distance metric related to both the class labels of the prototypes and the query patterns is proposed. Compared with the existing distance metrics, the proposed metric...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2006